Semantic Answer Validation in Question Answering Systems for Reading Comprehension Tests
نویسندگان
چکیده
In this paper it is presented a methodology for tackling the problem of answer validation in question answering for reading comprehension tests. The implemented system accepts a document as input and it answers multiple choice questions about it based on semantic similarity measures. It uses the Lucene information retrieval engine for carrying out information extraction employing additional automated linguistic processing such as stemming, anaphora resolution and part-of-speech tagging. The proposed approach validates the answers, by comparing the text retrieved by Lucene for each question with respect to its candidate answers. For this purpose, a validation based on semantic similarity is executed. We have evaluated the experiments carried out in order to verify the quality of the methodology proposed using a corpus widely used in international forums. The obtained results show that the proposed system selects the correct answer to a given question with a percentage of 12% more than with a lexical similarity based validation.
منابع مشابه
A Rule-Based Question Answering System For Reading Comprehension Tests
We have developed a rule-based system, Quarc, that can reada short story and find the sentence in the story that best answers a given question. Quarc uses heuristic rules that look for lexical and semantic clues in the question and the story. We have tested Quarc on reading comprehension tests typically given to children in grades 3-6. Overall, Quarc found the correct sentence 40% of the time, ...
متن کاملCloze validation against IELTS Reading Paper: Doubts on correlational validation
Cloze was officially introduced in a journal on Journalism as a technique for estimating text readability and as "a new psychological tool for measuring the effectiveness of communication" (Taylor, 1953: 415). Different varieties of cloze have since been developed and experimented upon as measures of such diverse traits as reading comprehension and language proficiency. The findings of numerous...
متن کاملAdversarial Examples for Evaluating Reading Comprehension Systems
Standard accuracy metrics indicate that reading comprehension systems are making rapid progress, but the extent to which these systems truly understand language remains unclear. To reward systems with real language understanding abilities, we propose an adversarial evaluation scheme for the Stanford Question Answering Dataset (SQuAD). Our method tests whether systems can answer questions about ...
متن کاملCSGS: Adapting a Short Answer Scoring System for Multiple-choice Reading Comprehension Exercises
This paper describes our system submission to the CLEF Question Answering Track 2014 Entrance Exam shared task competition, where the task is to correctly answer multiple choice reading comprehension exercises. Our system is a straightforward adaptation of a model originally designed for scoring short answers given by language learners to reading comprehension questions. Our model implements a ...
متن کاملGenerating Annotated Corpora for Reading Comprehension and Question Answering Evaluation
Recently, reading comprehension tests for students and adult language learners have received increased attention within the NLP community as a means to develop and evaluate robust question answering (NLQA) methods. We present our ongoing work on automatically creating richly annotated corpus resources for NLQA and on comparing automatic methods for answering questions against this data set. Sta...
متن کامل